Mixed-Strategy Learning With Continuous Action Sets
نویسندگان
چکیده
منابع مشابه
Game-theoretical control with continuous action sets
Motivated by the recent applications of game-theoretical learning techniques to the design of distributed control systems, we study a class of control problems that can be formulated as potential games with continuous action sets, and we propose an actor-critic reinforcement learning algorithm that provably converges to equilibrium in this class of problems. The method employed is to analyse th...
متن کاملStochastic fictitious play with continuous action sets
Continuous action space games form a natural extension to normal form games with finite action sets. However, whilst learning dynamics in normal form games are now well studied, it is not until recently that their continuous action space counterparts have been examined. We extend stochastic fictitious play to the continuous action space framework. In normal form games the limiting behaviour of ...
متن کاملLearning Continuous Action Models in a Real-Time Strategy Envir
Although several researchers have integrated methods for reinforcement learning (RL) with case-based reasoning (CBR) to model continuous action spaces, existing integrations typically employ discrete approximations of these models. This limits the set of actions that can be modeled, and may lead to non-optimal solutions. We introduce the Continuous Action and State Space Learner (CASSL), an int...
متن کاملLearning Continuous Action Models in a Real-Time Strategy Environment
Although several researchers have integrated methods for reinforcement learning (RL) with case-based reasoning (CBR) to model continuous action spaces, existing integrations typically employ discrete approximations of these models. This limits the set of actions that can be modeled, and may lead to non-optimal solutions. We introduce the Continuous Action and State Space Learner (CASSL), an int...
متن کاملAction learning versus strategy learning
This paper seeks to ascertain whether the strategy-learning model of Hanaki, Sethi, Erev, and Peterhansl (2003) better accounts for observed behavior than do the various action-learning models. It does so by measuring the goodness-of-fit of the models’ predictions against published experimental results for such games as Coordination, Prisoner’s Dilemma, and Chicken. The fit is measured via the ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Transactions on Automatic Control
سال: 2017
ISSN: 0018-9286,1558-2523,2334-3303
DOI: 10.1109/tac.2015.2511930